Search results for "Random subspace method"

showing 10 items of 11 documents

Diversity in search strategies for ensemble feature selection

2005

Ensembles of learnt models constitute one of the main current directions in machine learning and data mining. Ensembles allow us to achieve higher accuracy, which is often not achievable with single models. It was shown theoretically and experimentally that in order for an ensemble to be effective, it should consist of base classifiers that have diversity in their predictions. One technique, which proved to be effective for constructing an ensemble of diverse base classifiers, is the use of different feature subsets, or so-called ensemble feature selection. Many ensemble feature selection strategies incorporate diversity as an objective in the search for the best collection of feature subse…

business.industryContext (language use)Feature selectionMachine learningcomputer.software_genreEnsemble learningMeasure (mathematics)Random subspace methodEnsembles of classifiersComputingMethodologies_PATTERNRECOGNITIONHardware and ArchitectureFeature (computer vision)Signal ProcessingArtificial intelligenceData miningbusinesscomputerSoftwareSelection (genetic algorithm)Information SystemsMathematics
researchProduct

A local complexity based combination method for decision forests trained with high-dimensional data

2012

Accurate machine learning with high-dimensional data is affected by phenomena known as the “curse” of dimensionality. One of the main strategies explored in the last decade to deal with this problem is the use of multi-classifier systems. Several of such approaches are inspired by the Random Subspace Method for the construction of decision forests. Furthermore, other studies rely on estimations of the individual classifiers' competence, to enhance the combination in the multi-classifier and improve the accuracy. We propose a competence estimate which is based on local complexity measurements, to perform a weighted average combination of the decision forest. Experimental results show how thi…

Clustering high-dimensional dataComputational complexity theorybusiness.industryComputer scienceDecision treeMachine learningcomputer.software_genreRandom forestRandom subspace methodArtificial intelligenceData miningbusinessCompetence (human resources)computerClassifier (UML)Curse of dimensionality2012 12th International Conference on Intelligent Systems Design and Applications (ISDA)
researchProduct

Decision Committee Learning with Dynamic Integration of Classifiers

2000

Decision committee learning has demonstrated spectacular success in reducing classification error from learned classifiers. These techniques develop a classifier in the form of a committee of subsidiary classifiers. The combination of outputs is usually performed by majority vote. Voting, however, has a shortcoming. It is unable to take into account local expertise. When a new instance is difficult to classify, then the average classifier will give a wrong prediction, and the majority vote will more probably result in a wrong prediction. Instead of voting, dynamic integration of classifiers can be used, which is based on the assumption that each committee member is best inside certain subar…

Majority ruleBoosting (machine learning)business.industryComputer scienceFeature vectormedia_common.quotation_subjectMachine learningcomputer.software_genreRandom subspace methodComputingMethodologies_PATTERNRECOGNITIONVotingArtificial intelligenceAdaBoostbusinesscomputerClassifier (UML)Information integrationmedia_common
researchProduct

Ensemble Feature Selection Based on Contextual Merit and Correlation Heuristics

2001

Recent research has proven the benefits of using ensembles of classifiers for classification problems. Ensembles of diverse and accurate base classifiers are constructed by machine learning methods manipulating the training sets. One way to manipulate the training set is to use feature selection heuristics generating the base classifiers. In this paper we examine two of them: correlation-based and contextual merit -based heuristics. Both rely on quite similar assumptions concerning heterogeneous classification problems. Experiments are considered on several data sets from UCI Repository. We construct fixed number of base classifiers over selected feature subsets and refine the ensemble iter…

Training setbusiness.industryComputer scienceFeature selectionPattern recognitionBase (topology)Machine learningcomputer.software_genreExpert systemRandom subspace methodComputingMethodologies_PATTERNRECOGNITIONEnsembles of classifiersFeature (machine learning)Artificial intelligencebusinessHeuristicscomputerCascading classifiers
researchProduct

Ranking of Brain Tumour Classifiers Using a Bayesian Approach

2009

This study presents a ranking for classifers using a Bayesian perspective. This ranking framework is able to evaluate the performance of the models to be compared when they are inferred from different sets of data. It also takes into account the performance obtained on samples not used during the training of the classifiers. Besides, this ranking assigns a prior to each model based on a measure of similarity of the training data to a test case. An evaluation consisting of ranking brain tumour classifiers is presented. These multilayer perceptron classifiers are trained with 1H magnetic resonance spectroscopy (MRS) signals following a multiproject multicenter evaluation approach. We demonstr…

Measure (data warehouse)Training setComputer sciencebusiness.industryPerspective (graphical)Bayesian probabilityPattern recognitionMachine learningcomputer.software_genreRanking (information retrieval)Random subspace methodSimilarity (network science)Multilayer perceptronArtificial intelligencebusinesscomputer
researchProduct

Feature Selection for Ensembles of Simple Bayesian Classifiers

2002

A popular method for creating an accurate classifier from a set of training data is to train several classifiers, and then to combine their predictions. The ensembles of simple Bayesian classifiers have traditionally not been a focus of research. However, the simple Bayesian classifier has much broader applicability than previously thought. Besides its high classification accuracy, it also has advantages in terms of simplicity, learning speed, classification speed, storage space, and incrementality. One way to generate an ensemble of simple Bayesian classifiers is to use different feature subsets as in the random subspace method. In this paper we present a technique for building ensembles o…

Training setComputer sciencebusiness.industryBayesian probabilityPattern recognitionFeature selectionMachine learningcomputer.software_genreLinear subspaceRandom subspace methodNaive Bayes classifierComputingMethodologies_PATTERNRECOGNITIONIterative refinementArtificial intelligencebusinesscomputerClassifier (UML)Cascading classifiers
researchProduct

Evaluating Classifiers for Mobile-Masquerader Detection

2006

As a result of the impersonation of a user of a mobile terminal, sensitive information kept locally or accessible over the network can be abused. The means of masquerader detection are therefore needed to detect the cases of impersonation. In this paper, the problem of mobile-masquerader detection is considered as a problem of classifying the user behaviour as originating from the legitimate user or someone else. Different behavioural characteristics are analysed by designated one-class classifiers whose classifications are combined. The paper focuses on selecting the classifiers for mobile-masquerader detection. The selection process is conducted in two phases. First, the classification ac…

business.industryComputer scienceSmall numberLinear classifierPattern recognitionMachine learningcomputer.software_genreRandom subspace methodInformation sensitivityComputingMethodologies_PATTERNRECOGNITIONArtificial intelligencebusinesscomputerClassifier (UML)
researchProduct

Bagging and Boosting with Dynamic Integration of Classifiers

2000

One approach in classification tasks is to use machine learning techniques to derive classifiers using learning instances. The co-operation of several base classifiers as a decision committee has succeeded to reduce classification error. The main current decision committee learning approaches boosting and bagging use resampling with the training set and they can be used with different machine learning techniques which derive base classifiers. Boosting uses a kind of weighted voting and bagging uses equal weight voting as a combining method. Both do not take into account the local aspects that the base classifiers may have inside the problem space. We have proposed a dynamic integration tech…

Boosting (machine learning)Training setbusiness.industryComputer sciencemedia_common.quotation_subjectWeighted votingMachine learningcomputer.software_genreBoosting methods for object categorizationRandom subspace methodComputingMethodologies_PATTERNRECOGNITIONEnsembles of classifiersVotingAdaBoostArtificial intelligenceGradient boostingbusinesscomputermedia_common
researchProduct

Ensemble feature selection with the simple Bayesian classification

2003

Abstract A popular method for creating an accurate classifier from a set of training data is to build several classifiers, and then to combine their predictions. The ensembles of simple Bayesian classifiers have traditionally not been a focus of research. One way to generate an ensemble of accurate and diverse simple Bayesian classifiers is to use different feature subsets generated with the random subspace method. In this case, the ensemble consists of multiple classifiers constructed by randomly selecting feature subsets, that is, classifiers constructed in randomly chosen subspaces. In this paper, we present an algorithm for building ensembles of simple Bayesian classifiers in random sub…

business.industryBayesian probabilityFeature selectionPattern recognitionMachine learningcomputer.software_genreLinear subspaceRandom subspace methodNaive Bayes classifierBayes' theoremComputingMethodologies_PATTERNRECOGNITIONHardware and ArchitectureSignal ProcessingArtificial intelligencebusinesscomputerClassifier (UML)SoftwareCascading classifiersInformation SystemsMathematicsInformation Fusion
researchProduct

Dynamic integration of classifiers in the space of principal components

2003

Recent research has shown the integration of multiple classifiers to be one of the most important directions in machine learning and data mining. It was shown that, for an ensemble to be successful, it should consist of accurate and diverse base classifiers. However, it is also important that the integration procedure in the ensemble should properly utilize the ensemble diversity. In this paper, we present an algorithm for the dynamic integration of classifiers in the space of extracted features (FEDIC). It is based on the technique of dynamic integration, in which local accuracy estimates are calculated for each base classifier of an ensemble, in the neighborhood of a new instance to be pr…

Random subspace methodInformation extractionComputingMethodologies_PATTERNRECOGNITIONComputer sciencePrincipal component analysisFeature extractionData miningcomputer.software_genrecomputerClassifier (UML)Numerical integrationInformation integrationCurse of dimensionality
researchProduct